在本文中,我们提出了一种一般稳健的子带自适应滤波(GR-SAF)方案,以防止冲动噪声,通过在随机步行模型下以各个重量不确定性最小化均方根偏差。具体而言,通过选择不同的缩放因子,例如在GR-SAF方案中从M-估计和最大correntropy robust标准中选择,我们可以轻松获得不同的GR-SAF算法。重要的是,所提出的GR-SAF算法可以简化为可变的正则化鲁棒归一化的SAF算法,从而具有快速的收敛速率和低稳态误差。在系统识别的背景下,用冲动噪声和回声取消进行双词的模拟已证实,所提出的GR-SAF算法的表现优于其对应物。
translated by 谷歌翻译
最近的研究利用稀疏的分类来预测高维大脑活动信号的分类变量,以暴露人类的意图和精神状态,从而自动选择模型训练过程中的相关特征。但是,现有的稀疏分类模型可能会容易出现由大脑记录固有的噪声引起的性能降解。为了解决这个问题,我们旨在在本研究中提出一种新的健壮和稀疏分类算法。为此,我们将CorrentRopy学习框架引入基于自动相关性的稀疏分类模型,并提出了一种新的基于Correntropy的鲁棒稀疏逻辑回归算法。为了证明所提出算法的上等大脑活性解码性能,我们在合成数据集,脑电图(EEG)数据集和功能磁共振成像(FMRI)数据集上对其进行了评估。广泛的实验结果证实,不仅提出的方法可以在嘈杂和高维分类任务中实现更高的分类精度,而且还将为解码方案选择那些更有信息的功能。将Correntropy学习方法与自动相关性测定技术相结合,将显着提高噪声的鲁棒性,从而导致更足够的稳健稀疏脑解码算法。它在现实世界中的大脑活动解码和脑部计算机界面中提供了一种更强大的方法。
translated by 谷歌翻译
在实际人群计算应用程序中,图像中的人群密度差异很大。当面对密度变化时,人类倾向于在低密度区域定位和计数目标,并推理高密度区域的数量。我们观察到,CNN使用固定大小的卷积内核专注于局部信息相关性,而变压器可以通过使用全球自我注意机制有效地提取语义人群信息。因此,CNN可以在低密度区域中准确定位和估计人群,而在高密度区域中很难正确感知密度。相反,变压器在高密度区域具有很高的可靠性,但未能在稀疏区域定位目标。 CNN或变压器都无法很好地处理这种密度变化。为了解决此问题,我们提出了一个CNN和变压器自适应选择网络(CTASNET),该网络可以自适应地为不同密度区域选择适当的计数分支。首先,CTASNET生成CNN和变压器的预测结果。然后,考虑到CNN/变压器适用于低/高密度区域,密度引导的自适应选择模块被设计为自动结合CNN和Transformer的预测。此外,为了减少注释噪声的影响,我们引入了基于Correntropy的最佳运输损失。对四个挑战的人群计数数据集进行了广泛的实验,已经验证了该方法。
translated by 谷歌翻译
背景噪声和规模变化是人群计数中长期以来已经认识到的常见问题。人类瞥见人群的形象,立即知道人类的大概数量,以及他们通过关注的人群地区和人群地区的拥塞程度,并具有全球接收领域。因此,在本文中,我们通过对人类自上而下的视觉感知机制进行建模,提出了一个具有称为RANET的区域感知块的新型反馈网络。首先,我们介绍了一个反馈体系结构,以生成优先级地图,这些图提供了输入图像中候选人人群区域的先验。先验使Ranet更加关注人群地区。然后,我们设计了可以通过全局接受字段自适应地将上下文信息编码为输入图像的区域感知块。更具体地说,我们以列向量的形式扫描整个输入图像及其优先级图,以获得相关矩阵估计其相似性。获得的相关矩阵将用于建立像素之间的全球关系。我们的方法在几个公共数据集上优于最先进的人群计数方法。
translated by 谷歌翻译
提出了一种称为误差损失网络(ELN)的新型模型,以构建监督学习的误差损失函数。 ELN的结构类似于径向基函数(RBF)神经网络,但其输入是误差样本,输出是与该误差样本相对应的损耗。这意味着ELN的非线性输入输出映射器会创建误差损失函数。拟议的ELN为大量错误损失函数提供了统一模型,其中包括一些信息理论学习(ITL)损失函数作为特殊情况。 ELN的激活函数,权重参数和网络大小可以从误差样本中进行预先确定或学到。在此基础上,我们提出了一个新的机器学习范式,其中学习过程分为两个阶段:首先,使用ELN学习损失函数;其次,使用学习的损失功能继续执行学习。提出了实验结果,以证明新方法的理想性能。
translated by 谷歌翻译
近年来,守则已经安全地应用于强大的自适应过滤,以消除脉冲噪声或异常值的不利影响。正文通常被定义为两个随机变量之间的高斯内核的期望。当两个随机变量之间的误差对称地分布零点时,此定义是合理的。对于不对称错误分布的情况,对称高斯内核不合适,并且无法适应错误分布。为了解决这个问题,在这篇简短的情况下,我们提出了一种新的正文变异,名称不对称的正文,它使用非对称高斯模型作为内核功能。此外,开发了一种基于非对称控制的鲁棒自适应滤波算法,分析了其稳态收敛性能。提供了模拟以确认所提出的算法的理论结果和良好性能。
translated by 谷歌翻译
Deep learning models can achieve high accuracy when trained on large amounts of labeled data. However, real-world scenarios often involve several challenges: Training data may become available in installments, may originate from multiple different domains, and may not contain labels for training. Certain settings, for instance medical applications, often involve further restrictions that prohibit retention of previously seen data due to privacy regulations. In this work, to address such challenges, we study unsupervised segmentation in continual learning scenarios that involve domain shift. To that end, we introduce GarDA (Generative Appearance Replay for continual Domain Adaptation), a generative-replay based approach that can adapt a segmentation model sequentially to new domains with unlabeled data. In contrast to single-step unsupervised domain adaptation (UDA), continual adaptation to a sequence of domains enables leveraging and consolidation of information from multiple domains. Unlike previous approaches in incremental UDA, our method does not require access to previously seen data, making it applicable in many practical scenarios. We evaluate GarDA on two datasets with different organs and modalities, where it substantially outperforms existing techniques.
translated by 谷歌翻译
The development of social media user stance detection and bot detection methods rely heavily on large-scale and high-quality benchmarks. However, in addition to low annotation quality, existing benchmarks generally have incomplete user relationships, suppressing graph-based account detection research. To address these issues, we propose a Multi-Relational Graph-Based Twitter Account Detection Benchmark (MGTAB), the first standardized graph-based benchmark for account detection. To our knowledge, MGTAB was built based on the largest original data in the field, with over 1.55 million users and 130 million tweets. MGTAB contains 10,199 expert-annotated users and 7 types of relationships, ensuring high-quality annotation and diversified relations. In MGTAB, we extracted the 20 user property features with the greatest information gain and user tweet features as the user features. In addition, we performed a thorough evaluation of MGTAB and other public datasets. Our experiments found that graph-based approaches are generally more effective than feature-based approaches and perform better when introducing multiple relations. By analyzing experiment results, we identify effective approaches for account detection and provide potential future research directions in this field. Our benchmark and standardized evaluation procedures are freely available at: https://github.com/GraphDetec/MGTAB.
translated by 谷歌翻译
As one of the prevalent methods to achieve automation systems, Imitation Learning (IL) presents a promising performance in a wide range of domains. However, despite the considerable improvement in policy performance, the corresponding research on the explainability of IL models is still limited. Inspired by the recent approaches in explainable artificial intelligence methods, we proposed a model-agnostic explaining framework for IL models called R2RISE. R2RISE aims to explain the overall policy performance with respect to the frames in demonstrations. It iteratively retrains the black-box IL model from the randomized masked demonstrations and uses the conventional evaluation outcome environment returns as the coefficient to build an importance map. We also conducted experiments to investigate three major questions concerning frames' importance equality, the effectiveness of the importance map, and connections between importance maps from different IL models. The result shows that R2RISE successfully distinguishes important frames from the demonstrations.
translated by 谷歌翻译
Compressed videos often exhibit visually annoying artifacts, known as Perceivable Encoding Artifacts (PEAs), which dramatically degrade video visual quality. Subjective and objective measures capable of identifying and quantifying various types of PEAs are critical in improving visual quality. In this paper, we investigate the influence of four spatial PEAs (i.e. blurring, blocking, bleeding, and ringing) and two temporal PEAs (i.e. flickering and floating) on video quality. For spatial artifacts, we propose a visual saliency model with a low computational cost and higher consistency with human visual perception. In terms of temporal artifacts, self-attention based TimeSFormer is improved to detect temporal artifacts. Based on the six types of PEAs, a quality metric called Saliency-Aware Spatio-Temporal Artifacts Measurement (SSTAM) is proposed. Experimental results demonstrate that the proposed method outperforms state-of-the-art metrics. We believe that SSTAM will be beneficial for optimizing video coding techniques.
translated by 谷歌翻译